social media post
British athletes given AI app as shield from online abuse
Team GB Olympic and Paralympic athletes are being offered a new form of artificial intelligence-based protection from online abuse. UK Sport, the body that funds Olympic and Paralympic sports, has signed a contract worth more than £300,000 to give thousands of athletes access to an app that detects and hides abusive posts sent by other users on social media. Athletes are able to sign up for free and can protect their accounts throughout the Games cycle up to Los Angeles 2028. The level of abuse our athletes are facing online is unacceptable - to do nothing about this is not an option, UK Sport director of performance Kate Baker said about a deal that is the first of its kind in British sport. The app, called Social Protect, uses AI to try to ensure athletes see as few abusive messages sent their way as possible.
- North America > United States > California > Los Angeles County > Los Angeles (0.25)
- Europe > United Kingdom > England > Tyne and Wear > Sunderland (0.06)
- Oceania > Australia (0.05)
- (4 more...)
OPOR-Bench: Evaluating Large Language Models on Online Public Opinion Report Generation
Yu, Jinzheng, Xu, Yang, Li, Haozhen, Li, Junqi, Feng, Yifan, Zhu, Ligu, Shen, Hao, Shi, Lei
Online Public Opinion Reports consolidate news and social media for timely crisis management by governments and enterprises. While large language models have made automated report generation technically feasible, systematic research in this specific area remains notably absent, particularly lacking formal task definitions and corresponding benchmarks. To bridge this gap, we define the Automated Online Public Opinion Report Generation (OPOR-GEN) task and construct OPOR-BENCH, an event-centric dataset covering 463 crisis events with their corresponding news articles, social media posts, and a reference summary. To evaluate report quality, we propose OPOR-EVAL, a novel agent-based framework that simulates human expert evaluation by analyzing generated reports in context. Experiments with frontier models demonstrate that our framework achieves high correlation with human judgments. Our comprehensive task definition, benchmark dataset, and evaluation framework provide a solid foundation for future research in this critical domain.
- Europe > Belgium > Brussels-Capital Region > Brussels (0.04)
- Asia > China > Heilongjiang Province > Harbin (0.04)
- Asia > China > Beijing > Beijing (0.04)
- (9 more...)
- Media > News (1.00)
- Health & Medicine > Therapeutic Area (0.93)
- Information Technology (0.93)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Extracting Disaster Impacts and Impact Related Locations in Social Media Posts Using Large Language Models
Hameed, Sameeah Noreen, Ranathunga, Surangika, Prasanna, Raj, Stock, Kristin, Jones, Christopher B.
Large-scale disasters can often result in catastrophic consequences on people and infrastructure. Situation awareness about such disaster impacts generated by authoritative data from in-situ sensors, remote sensing imagery, and/or geographic data is often limited due to atmospheric opacity, satellite revisits, and time limitations. This often results in geo-temporal information gaps. In contrast, impact-related social media posts can act as "geo-sensors" during a disaster, where people describe specific impacts and locations. However, not all locations mentioned in disaster-related social media posts relate to an impact. Only the impacted locations are critical for directing resources effectively. e.g., "The death toll from a fire which ripped through the Greek coastal town of #Mati stood at 80, with dozens of people unaccounted for as forensic experts tried to identify victims who were burned alive #Greecefires #AthensFires #Athens #Greece." contains impacted location "Mati" and non-impacted locations "Greece" and "Athens". This research uses Large Language Models (LLMs) to identify all locations, impacts and impacted locations mentioned in disaster-related social media posts. In the process, LLMs are fine-tuned to identify only impacts and impacted locations (as distinct from other, non-impacted locations), including locations mentioned in informal expressions, abbreviations, and short forms. Our fine-tuned model demonstrates efficacy, achieving an F1-score of 0.69 for impact and 0.74 for impacted location extraction, substantially outperforming the pre-trained baseline. These robust results confirm the potential of fine-tuned language models to offer a scalable solution for timely decision-making in resource allocation, situational awareness, and post-disaster recovery planning for responders.
- Europe > Greece > Attica > Athens (0.24)
- North America > Haiti (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- (21 more...)
- Health & Medicine (1.00)
- Information Technology > Services (0.67)
- Government > Military (0.54)
- Energy > Renewable > Geothermal > Geothermal Energy Exploration and Development > Geophysical Analysis & Survey (0.34)
BanglaMM-Disaster: A Multimodal Transformer-Based Deep Learning Framework for Multiclass Disaster Classification in Bangla
Islam, Ariful, Hossen, Md Rifat, Arif, Md. Mahmudul, Noman, Abdullah Al, Rahman, Md Arifur
Natural disasters remain a major challenge for Bangladesh, so real-time monitoring and quick response systems are essential. In this study, we present BanglaMM-Disaster, an end-to-end deep learning-based multimodal framework for disaster classification in Bangla, using both textual and visual data from social media. We constructed a new dataset of 5,037 Bangla social media posts, each consisting of a caption and a corresponding image, annotated into one of nine disaster-related categories. The proposed model integrates transformer-based text encoders, including BanglaBERT, mBERT, and XLM-RoBERTa, with CNN backbones such as ResNet50, DenseNet169, and MobileNetV2, to process the two modalities. Using early fusion, the best model achieves 83.76% accuracy. This surpasses the best text-only baseline by 3.84% and the image-only baseline by 16.91%. Our analysis also shows reduced misclassification across all classes, with noticeable improvements for ambiguous examples. This work fills a key gap in Bangla multimodal disaster analysis and demonstrates the benefits of combining multiple data types for real-time disaster response in low-resource settings.
- Asia > Bangladesh (0.26)
- North America > United States > Delaware > New Castle County > New Castle (0.04)
- Africa > Angola (0.04)
Reasoning-Guided Claim Normalization for Noisy Multilingual Social Media Posts
Sharma, Manan, Suneesh, Arya, Jain, Manish, Rajpoot, Pawan Kumar, Devadiga, Prasanna, Hazarika, Bharatdeep, Shrivastava, Ashish, Gurumurthy, Kishan, Suresh, Anshuman B, Baliga, Aditya U
We address claim normalization for multilingual misinformation detection - transforming noisy social media posts into clear, verifiable statements across 20 languages. The key contribution demonstrates how systematic decomposition of posts using Who, What, Where, When, Why and How questions enables robust cross-lingual transfer despite training exclusively on English data. Our methodology incorporates finetuning Qwen3-14B using LoRA with the provided dataset after intra-post deduplication, token-level recall filtering for semantic alignment and retrieval-augmented few-shot learning with contextual examples during inference. Our system achieves METEOR scores ranging from 41.16 (English) to 15.21 (Marathi), securing third rank on the English leaderboard and fourth rank for Dutch and Punjabi. The approach shows 41.3% relative improvement in METEOR over baseline configurations and substantial gains over existing methods. Results demonstrate effective cross-lingual generalization for Romance and Germanic languages while maintaining semantic coherence across diverse linguistic structures.
- Europe > Spain > Galicia > Madrid (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > Washington > King County > Seattle (0.04)
- (13 more...)
P-ReMIS: Pragmatic Reasoning in Mental Health and a Social Implication
Oram, Sneha, Bhattacharyya, Pushpak
Although explainability and interpretability have received significant attention in artificial intelligence (AI) and natural language processing (NLP) for mental health, reasoning has not been examined in the same depth. Addressing this gap is essential to bridge NLP and mental health through interpretable and reasoning-capable AI systems. To this end, we investigate the pragmatic reasoning capability of large-language models (LLMs) in the mental health domain. We introduce PRiMH dataset, and propose pragmatic reasoning tasks in mental health with pragmatic implicature and presupposition phenomena. In particular, we formulate two tasks in implicature and one task in presupposition. To benchmark the dataset and the tasks presented, we consider four models: Llama3.1, Mistral, MentaLLaMa, and Qwen. The results of the experiments suggest that Mistral and Qwen show substantial reasoning abilities in the domain. Subsequently, we study the behavior of MentaLLaMA on the proposed reasoning tasks with the rollout attention mechanism. In addition, we also propose three StiPRompts to study the stigma around mental health with the state-of-the-art LLMs, GPT4o-mini, Deepseek-chat, and Claude-3.5-haiku. Our evaluated findings show that Claude-3.5-haiku deals with stigma more responsibly compared to the other two LLMs.
- North America > United States (0.04)
- Europe > France > Provence-Alpes-Côte d'Azur > Bouches-du-Rhône > Marseille (0.04)
AI Models Get Brain Rot, Too
A new study shows that feeding large language models low-quality, high-engagement content from social media lowers their cognitive abilities. AI models may be a bit like humans, after all. A new study from the University of Texas at Austin, Texas A&M, and Purdue University shows that large language models fed a diet of popular but low-quality social media content experience a kind of "brain rot" that may be familiar to anyone who has spent too long doomscrolling on X or TikTok. We live in an age where information grows faster than attention spans--and much of it is engineered to capture clicks, not convey truth or depth," says Junyuan Hong, an incoming assistant professor at the National University of Singapore who worked on the study as a graduate student at UT Austin. "We wondered: What happens when AIs are trained on the same stuff?"
- North America > United States > Texas > Travis County > Austin (0.55)
- Asia > Singapore (0.25)
- North America > United States > California (0.05)
- (3 more...)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (0.86)
- Health & Medicine > Therapeutic Area > Neurology (0.72)
Natural Language Tools: A Natural Language Approach to Tool Calling In Large Language Agents
Johnson, Reid T., Pain, Michelle D., West, Jordan D.
We present Natural Language Tools (NLT), a framework that replaces programmatic JSON tool calling in large language models (LLMs) with natural language outputs. By decoupling tool selection from response generation, NLT eliminates task interference and format constraints that degrade tool call performance. When evaluated across 10 models and 6,400 trials spanning customer service and mental health domains, NLT improves tool calling accuracy by 18.4 percentage points while reducing output variance by 70%. Open-weight models see the largest gains, surpassing flagship closed-weight alternatives, with implications for model training in both reinforcement learning and supervised fine-tuning stages. These improvements persist under prompt perturbations and extend tool-calling capabilities to models lacking native support.
- North America > Canada (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- Asia > Middle East > Jordan (0.04)
TIFIN India at SemEval-2025: Harnessing Translation to Overcome Multilingual IR Challenges in Fact-Checked Claim Retrieval
Devadiga, Prasanna, Suneesh, Arya, Rajpoot, Pawan Kumar, Hazarika, Bharatdeep, Baliga, Aditya U
We address the challenge of retrieving previously fact-checked claims in monolingual and crosslingual settings - a critical task given the global prevalence of disinformation. Our approach follows a two-stage strategy: a reliable baseline retrieval system using a fine-tuned embedding model and an LLM-based reranker. Our key contribution is demonstrating how LLM-based translation can overcome the hurdles of multilingual information retrieval. Additionally, we focus on ensuring that the bulk of the pipeline can be replicated on a consumer GPU. Our final integrated system achieved a success@10 score of 0.938 and 0.81025 on the monolingual and crosslingual test sets, respectively.
- Asia > India (0.41)
- Europe > Austria > Vienna (0.14)
- North America > United States > Washington > King County > Seattle (0.04)
- (5 more...)
- Media > News (0.49)
- Information Technology (0.47)
AttentionDep: Domain-Aware Attention for Explainable Depression Severity Assessment
Ibrahimov, Yusif, Anwar, Tarique, Yuan, Tommy, Mutallimov, Turan, Hasanov, Elgun
Abstract-- In today's interconnected society, social media platforms provide a window into individuals' thoughts, emotions, and mental states. This paper explores the use of platforms like Facebook, X (formerly Twitter), and Reddit for depression severity detection. We propose AttentionDep, a domain-aware attention model that drives explainable depression severity estimation by fusing contextual and domain knowledge. Posts are encoded hierarchically using unigrams and bigrams, with attention mechanisms highlighting clinically relevant tokens. Domain knowledge from a curated mental health knowledge graph is incorporated through a cross-attention mechanism, enriching the contextual features. Finally, depression severity is predicted using an ordinal regression framework that respects the clinical-relevance and natural ordering of severity levels. Depression affects over 280 million people globally, with severe outcomes, including approximately 700,000 suicides annually [1]. Despite available treatments, barriers such as stigma and limited access of healthcare treatments leave over 70% of affected individuals untreated [2]. The COVID-19 pandemic has intensified this crisis, highlighting the urgent need for effective and scalable methods for early symptom identification [3]. Social media platforms such as Facebook, X, and Reddit provide a rich source of user-generated content reflecting mental states, offering opportunities for automated depression assessment [4], [5]. Traditional interview-and questionnaire-based approaches, while informative, are resource-intensive and may lack scalability [6], [7].
- Europe > France (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (11 more...)
- Research Report > New Finding (0.46)
- Research Report > Experimental Study (0.46)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (0.93)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.68)